SERAPH: Semi-supervised Metric Learning Paradigm with Hyper Sparsity
نویسندگان
چکیده
We consider the problem of learning a distance metric from a limited amount of pairwise information as effectively as possible. The proposed SERAPH (SEmi-supervised metRic leArning Paradigm with Hyper sparsity) is a direct and substantially more natural approach for semi-supervised metric learning, since the supervised and unsupervised parts are based on a unified information theoretic framework. Unlike other extensions, the unsupervised part of SERAPH can extract further pairwise information from the unlabeled data according to temporary results of the supervised part, and therefore interacts with the supervised part positively. SERAPH involves both the sparsity of posterior distributions over the unobserved weak labels and the sparsity of the induced projection matrices, which we call the hyper sparsity. The resulting optimization is solved by an EM-like scheme, where the M-Step is convex, and the E-Step has analytical solution. Experimental results show that SERAPH compares favorably with existing metric learning algorithms based on weak labels.
منابع مشابه
Information-theoretic Semi-supervised Metric Learning via Entropy Regularization
We propose a general information-theoretic approach to semi-supervised metric learning called SERAPH (SEmi-supervised metRic leArning Paradigm with Hypersparsity) that does not rely on the manifold assumption. Given the probability parameterized by a Mahalanobis distance, we maximize its entropy on labeled data and minimize its entropy on unlabeled data following entropy regularization. For met...
متن کاملComposite Kernel Optimization in Semi-Supervised Metric
Machine-learning solutions to classification, clustering and matching problems critically depend on the adopted metric, which in the past was selected heuristically. In the last decade, it has been demonstrated that an appropriate metric can be learnt from data, resulting in superior performance as compared with traditional metrics. This has recently stimulated a considerable interest in the to...
متن کاملSome Research Problems in Metric Learning and Manifold Learning
In the past few years, metric learning, semi-supervised learning, and manifold learning methods have aroused a great deal of interest in the machine learning community. Many machine learning and pattern recognition algorithms rely on a distance metric. Instead of choosing the metric manually, a promising approach is to learn the metric from data automatically. Besides some early work on metric ...
متن کاملNon-Negative Semi-Supervised Learning
The contributions of this paper are three-fold. First, we present a general formulation for reaping the benefits from both non-negative data factorization and semi-supervised learning, and the solution naturally possesses the characteristics of sparsity, robustness to partial occlusions, and greater discriminating power via extra unlabeled data. Then, an efficient multiplicative updating proced...
متن کاملAdversarial Dropout for Supervised and Semi-supervised Learning
Recently, training with adversarial examples, which are generated by adding a small but worst-case perturbation on input examples, has improved the generalization performance of neural networks. In contrast to the biased individual inputs to enhance the generality, this paper introduces adversarial dropout, which is a minimal set of dropouts that maximize the divergence between 1) the training ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- CoRR
دوره abs/1105.0167 شماره
صفحات -
تاریخ انتشار 2011